Starting engagement detection towards a companion robot using multimodal features

نویسندگان

  • Dominique Vaufreydaz
  • Wafa Johal
  • Claudine Combe
چکیده

Recognition of intentions is a subconscious cognitive process vital to human communication. This skill enables anticipation and increases the quality of interactions between humans. Within the context of engagement, non-verbal signals are used to communicate the intention of starting the interaction with a partner. In this paper, we investigated methods to detect these signals in order to allow a robot to know when it is about to be addressed. Originality of our approach resides in taking inspiration from social and cognitive sciences to perform our perception task. We investigate meaningful features, i.e. human readable features, and elicit which of these are important for recognizing someone’s intention of starting an interaction. Classically, spatial information like the human position and speed, the human-robot distance are used to detect the engagement. Our approach integrates multimodal features gathered using a companion robot equipped with a Kinect. The evaluation on our corpus collected in spontaneous conditions highlights its robustness and validates the use of such a technique in a real environment. Experimental validation shows that multimodal features set gives better precision and recall than using only spatial and speed features. We also demonstrate that 7 selected features are sufficient to provide a good starting engagement detection score. In our last investigation, we show that among our full 99 features set, the space reduction is not a solved task. This result opens new researches perspectives on multimodal engagement detection.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Towards an affect sensitive interactive companion

As robots are increasingly being viewed as social entities to be integrated in our daily lives, social perceptive abilities seem a necessary requirement for enabling more natural interaction with human users. In this paper, we present an interaction scenario where user play chess with an iCat robot and propose an affect recognition system that uses computational models to automatically extract ...

متن کامل

Social Robots in Learning Environments: a Case Study of an Empathic Chess Companion

We present a scenario where a social robot acts as a chess companion for children, and describe our current efforts towards endowing such robot with empathic capabilities. A multimodal framework for modeling some of the user’s affective states that combines visual and task-related features is presented. Further, we describe how the robot selects adaptive empathic responses considering the model...

متن کامل

Modelling Empathy in Social Robotic Companions

Empathy can be broadly defined as the ability to understand and respond appropriately to the affective states of others. In this paper, we present a scenario where a social robot acts as a chess companion for children, and describe our current efforts towards endowing such robot with empathic capabilities. A multimodal framework for modeling some of the user’s affective states that combines vis...

متن کامل

Analysis of Engagement and User Experience with a Laughter Responsive Social Robot

We explore the effect of laughter perception and response in terms of engagement in human-robot interaction. We designed two distinct experiments in which the robot has two modes: laughter responsive and laughter non-responsive. In responsive mode, the robot detects laughter using a multimodal real-time laughter detection module and invokes laughter as a backchannel to users accordingly. In non...

متن کامل

Speaker Dependency Analysis, Audiovisual Fusion Cues and a Multimodal BLSTM for Conversational Engagement Recognition

Conversational engagement is a multimodal phenomenon and an essential cue to assess both human-human and human-robot communication. Speaker-dependent and speaker-independent scenarios were addressed in our engagement study. Handcrafted audio-visual features were used. Fixed window sizes for feature fusion method were analysed. Novel dynamic window size selection and multimodal bi-directional lo...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • Robotics and Autonomous Systems

دوره 75  شماره 

صفحات  -

تاریخ انتشار 2016